NEUROEVOLUTIONARY ALGORITHMS FOR NEURAL NETWORKS GENERATING
نویسندگان
چکیده
Solving engineering problems using conventional neural networks requires long-term research on the choice of architecture and hyperparameters. A strong artificial intelligence would be devoid such shortcomings. Such is carried out a very wide range approaches: for example, biological (attempts to grow brain in laboratory conditions), hardware (creating processors) or software (using power ordinary CPUs GPUs). The goal work develop system that allow evolutionary approaches generate suitable solving problems. This called “neuroevolution”. purpose this also includes study features possible applicable strategies. object neuroevolutionary approach machine learning. subject strategies, coding methods organism’s genome. scientific novelty lies testing previously unused strategies generalization obtained systems “general intelligence”. simulating neuroevolution was created. specifics implementation were considered, algorithms justified, their explained. In order perform experiments, datasets created applying developed. It choose most optimal training parameters, find relationship between them, as well accuracy speed training. cannot said models implemented within directly bring us closer AI. They still lack own memory certain level complexity. For successful use, it necessary configure view input data some calculations outside model. However, future, can developed, with SNNs, use special equipment
منابع مشابه
Generating Neural Networks with Neural Networks
Hypernetworks are neural networks that generate weights for another neural network. We formulate the hypernetwork training objective as a compromise between accuracy and diversity, where the diversity takes into account trivial symmetry transformations of the target network. We explain how this simple formulation generalizes variational inference. We use multi-layered perceptrons to form the ma...
متن کاملAlgorithms for Neural Networks
Parallelizable optimization techniques are applied to the problem of learning in feedforward neural networks. In addition to having superior convergence properties, optimization techniques such as the PolakRibiere method are also significantly more efficient than the Backpropagation algorithm. These results are based on experiments performed on small boolean learning problems and the noisy real...
متن کاملSelf-generating Neural Networks
This review of recent advances conceringing growing and pruning neural networks focuses on three areas. Adaptive resonance theory networks automatically have an architecture whose size adjusts to its task. Networks that optimize resource allocation have been around nearly fifteen years and developments are still being made in growing and pruning strategies, particularly for on-line, real-time a...
متن کاملHYBRID ARTIFICIAL NEURAL NETWORKS BASED ON ACO-RPROP FOR GENERATING MULTIPLE SPECTRUM-COMPATIBLE ARTIFICIAL EARTHQUAKE RECORDS FOR SPECIFIED SITE GEOLOGY
The main objective of this paper is to use ant optimized neural networks to generate artificial earthquake records. In this regard, training accelerograms selected according to the site geology of recorder station and Wavelet Packet Transform (WPT) used to decompose these records. Then Artificial Neural Networks (ANN) optimized with Ant Colony Optimization and resilient Backpropagation algorith...
متن کاملCommunication Efficient Algorithms for Generating Massive Networks
Massive complex systems are prevalent throughout all of our lives, from various biological systems as the human genome to technological networks such as Facebook or Twitter. Rapid advances in technology allow us to gather more and more data that is connected to these systems. Analyzing and extracting this huge amount of information is a crucial task for a variety of scientific disciplines. A co...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Vìsnik Hmel?nic?kogo nacìonal?nogo unìversitetu
سال: 2022
ISSN: ['2307-5732']
DOI: https://doi.org/10.31891/2307-5732-2022-315-6-240-244